Goto

Collaborating Authors

 text generation model






MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers

Neural Information Processing Systems

As major progress is made in open-ended text generation, measuring how close machine-generated text is to human language remains a critical open problem. We introduce Mauve, a comparison measure for open-ended text generation, which directly compares the learnt distribution from a text generation model to the distribution of human-written text using divergence frontiers.




SeqPA TE: Differentially Private Text Generation via Knowledge Distillation

Neural Information Processing Systems

P A TE is a recent DP learning algorithm that achieves high utility with strong privacy protection on training samples. However, text generation models output tokens sequentially in a large output space; the classic P A TE algorithm is not customized for this setting.



MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers

Neural Information Processing Systems

As major progress is made in open-ended text generation, measuring how close machine-generated text is to human language remains a critical open problem. We introduce Mauve, a comparison measure for open-ended text generation, which directly compares the learnt distribution from a text generation model to the distribution of human-written text using divergence frontiers. Through an extensive empirical study on three open-ended generation tasks, we find that Mauve identifies known properties of generated text, scales naturally with model size, and correlates with human judgments, with fewer restrictions than existing distributional evaluation metrics.